A Universal Approximation Theorem for Mixture-of-Experts Models

نویسندگان

  • Hien Duy Nguyen
  • Luke R. Lloyd-Jones
  • Geoffrey J. McLachlan
چکیده

The mixture-of-experts (MoE) model is a popular neural network architecture for nonlinear regression and classification. The class of MoE mean functions is known to be uniformly convergent to any unknown target function, assuming that the target function is from a Sobolev space that is sufficiently differentiable and that the domain of estimation is a compact unit hypercube. We provide an alternative result, which shows that the class of MoE mean functions is dense in the class of all continuous functions over arbitrary compact domains of estimation. Our result can be viewed as a universal approximation theorem for MoE models. The theorem we present allows MoE users to be confident in applying such models for estimation when data arise from nonlinear and nondifferentiable generative processes.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Time Series Prediction using Mixtures of Experts

We consider the problem of prediction of stationary time series, using the architecture known as mixtures of experts (MEM). Here we suggest a mixture which blends several autoregressive models. This study focuses on some theoretical foundations of the prediction problem in this context. More precisely, it is demonstrated that this model is a universal approximator, with respect to learning the ...

متن کامل

Discontinuity-Sensitive Optimal Control Learning by Mixture of Experts

This paper proposes a discontinuity-sensitive approach to learn the solutions of parametric optimal control problems with high accuracy. Many tasks, ranging from model predictive control to reinforcement learning, may be solved by learning optimal solutions as a function of problem parameters. However, nonconvexity, discrete homotopy classes, and control switching cause discontinuity in the par...

متن کامل

The Basic Theorem and its Consequences

Let T be a compact Hausdorff topological space and let M denote an n–dimensional subspace of the space C(T ), the space of real–valued continuous functions on T and let the space be equipped with the uniform norm. Zukhovitskii [7] attributes the Basic Theorem to E.Ya.Remez and gives a proof by duality. He also gives a proof due to Shnirel’man, which uses Helly’s Theorem, now the paper obtains a...

متن کامل

Universal Approximation of Interval-valued Fuzzy Systems Based on Interval-valued Implications

It is firstly proved that the multi-input-single-output (MISO) fuzzy systems based on interval-valued $R$- and $S$-implications can approximate any continuous function defined on a compact set to arbitrary accuracy.  A formula to compute the lower upper bounds on the number  of interval-valued fuzzy sets needed to achieve a pre-specified approximation  accuracy for an arbitrary multivariate con...

متن کامل

Universal Approximation by Ridge Computational Models and Neural Networks: A Survey

Computational models made up of linear combinations of ridge basis functions, widely used in machine learning and artificial intelligence, are considered. For such models, the literature on the so-called “universal approximation property” is surveyed. Different approaches, proof techniques, and tools are examined.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Neural computation

دوره 28 12  شماره 

صفحات  -

تاریخ انتشار 2016